Overview

Bayesian distribution regression (BDR) is built on a few pieces of theory:

  1. Kernel mean embeddings
  2. Gaussian process regression (as the prior for the kernel mean embeddings)
  3. Landmark points for dimensionality reduction (advantages of this over FastFood? both are reducing the dimensionality of the feature embedding so that we can calualate the empirical mean)

This notebook builds the theory of Bayesian Distribution Regression piece by piece:

  1. Explore kernels from the kernlab package. We’ll later use the functions from this package to calculate mean embeddings for bags of observations \(\{x_j^i\}_{j=1}^{N_i}\).

Problem statement

We observe survey responses \((\mathbf{x}_i, y_i)\) for \(i \in S\), where \(\mathbf{x}_i \in \mathbb{R}^{V_s}\) is a vector of \(v \in (1, \dots V_s)\) covariates (i.e. age, gender, education, income) and \(y_i\) denotes who a respondent says they’re likely to support in an upcoming political election. \(S \subset (1, \dots, N)\) denotes the set of indicies in the population of size \(N\) that were observed in the survey.

We also observe a much larger set of \(V_u >> V_s\) covariates \(\mathbf{x}_i \in \mathbb{R}^{V_u}\), \(V_s \subset V_u\), for all \(i \in D\), where \(D \subset N\). The covariates \(v \in V_u \setminus V_s\) are highly predictive of \(y_i\).

We say that a survey respondent is “matched” to the population if \(i \in M\) where \(M = S \cap D\).

The goal here is to estimate \(\hat{y}_i \forall i \in D\). Usually to do this, predictive models are built using only the “matched” data, or using respondents \(i \in M\). However, we would like to leverage \(i \in S \setminus M\), for which we observe \(y_i\) and \(x_{i, 1}, \dots, x_{i, V_s}\) but not \(\mathbf{x}_{i, \mathbf{v}}\) for \(\mathbf{v} \in V_u \setminus V_s\).

Question:

Explore Kernels from kernlab package

## create a RBF kernel function with sigma hyper-parameter 0.05
rbf = rbfdot(sigma = 1)
## create artificial data set
x <- matrix(rnorm(60), 6, 10)
## compute kernel matrix
kx <- kernelMatrix(rbf, x)  ## k_12 is equivalent to exp(- sum((x[1,] - x[2,])^2))

Basic Gaussian process

A Gaussian process is a collection of random variables, any finite subset of which follows a multivariate normal distribution

prior: \(f \sim \mathcal{GP}(\mathbf{0}, k(x, x'))\)

We fix a finite set of \(d\) points \(\mathbf{s} = (s_1, \dots, s_d)\) in order to draw from the GP as a multivariate normal. Conditioning on \(s\), this becomes \(\mathbf{f} \sim \mathcal{N}(\mathbf{0}_d, \Sigma_d)\) where \(\Sigma_d\) is the empirical covariance matrix of our points \(s\).

We then ``observe" 5 points \(\{x_i, y_i\}_{i = 1}^n\). We assume that y is observed without noise such that \(y = f(x)\). In order to predict at a new point \(x_\star\), we condiditon on the observed points \(X\) and their associated outcomes \(Y\), and arrive at the closed-form posterior predictive distribution of \(f_\star\)

posterior: \(f_\star | x_\star, X, \mathbf{y} \sim \mathcal{N}(k_\star K^{-1}\mathbf{y}, k_{\star\star} - k_\star K^{-1}k_\star^T)\)

where \(K_{ij} = k(x_i, x_j)\) and \(k_\star = [k(x_1, x_\star), \dots, k(x_n, x_\star)]\) and \(k_{\star\star} = k(x_star, x_star)\)

# define function to calculate empirical covariance matrix based on Gaussian kernel with length-scale l
calcSigma = function(x1, x2, l = 1){
  Sigma = matrix(nrow = length(x1), ncol = length(x2))
  
  for(i in 1:length(x1)){
    for(j in 1:length(x2)){
      Sigma[i,j] = exp(-1/2 * ((x1[i] - x2[j])/l)^2)
    }
  }
  return(Sigma)
}
# define the points that we want to fix for the function draw
s = seq(-5, 5, by = 0.1)
# calculate the empirical covariance of the points (basically 0)
Sigma = calcSigma(s, s)
# draw 5 samples from the GP prior at the fixed points
x_prior = t(mvrnorm(n = 5, mu = rep(0, length(s)), Sigma = Sigma))
x_draws = data.frame(cbind(s, x_prior))
x_draws = melt(x_draws, id = 's', value.name = 'f(x)')
# plot draws from the prior
plot_gp_prior = ggplot(x_draws, aes(x = s, y = `f(x)`, color = variable)) + geom_line() +
  ggtitle("Draws from GP Prior") + theme_minimal()
# specify points that we observe
x_obs = c(-4, -2.5, -1, 0, 4)
y_obs = c(-2, 0, 1, 2, -1)
# calculate kernel matricies needed for posterior evaluation
K_xx = calcSigma(x_obs, x_obs)
K_xstarx = calcSigma(s, x_obs)
K_xxstar = calcSigma(x_obs, s)
K_xstarxstar = calcSigma(s, s)
# calculate posterior mean and variance
mu = K_xstarx %*% solve(K_xx) %*% y_obs
Sigma_star = K_xstarxstar - K_xstarx %*% solve(K_xx) %*% K_xxstar
# draw 5 samples from the posterior at fixed eval points s
x_post = t(mvrnorm(5, mu = mu, Sigma = Sigma_star))
x_post = data.frame(cbind(s, x_post))
x_post = melt(x_post, id = 's', value.name = 'f_star')
# calculate 95% confidence bands for posterior
f_star_covar = data.frame(cbind(s, ub = mu + 2 * sqrt(diag(Sigma_star)), lb = mu - 2 * sqrt(diag(Sigma_star))))
NaNs producedNaNs produced
# plot draws from posterior
plot_gp_post = ggplot() + 
  geom_ribbon(data = f_star_covar, aes(x = s, ymin = V3, ymax = V2), fill = "grey", alpha = 0.5) +
  geom_line(data = x_post, aes(x = s, y = f_star, color = variable)) +
  #geom_point(aes(x = x_obs, y = y_obs)) +
  ggtitle("Draws from GP Posterior") +
  theme_minimal()
  
grid.arrange(plot_gp_prior, plot_gp_post, nrow = 1)

Try out BDR on survey data

Goal: Estimate individual-level support using BDR on state-level support outcomes

Data

Undeclared level(s) 2, 3, 4 added in variable: densityUndeclared level(s) 2, 3, 4 added in variable: sdensityUndeclared level(s) 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95 added in variable: ageUndeclared level(s) 1, 2, 3, 4, 5, 6, 7 added in variable: hh1Undeclared level(s) 1, 2, 3, 4, 5, 6, 7 added in variable: hh3

Individual-level survey responses (n = nrow(data_sept18) from September 2018 Pew Research survey (https://www.people-press.org/dataset/september-2018-political-survey/).

Recode survey responses

Recode the survey responses - combine q7 (strong) and q8 (leaners) to get full support response

## support 
data_sept18[, .N, .(q7, q8)]
data_sept18[, qsupport := NULL]
Adding new column 'qsupport' then assigning NULL (deleting it).
data_sept18[q7 == "Democratic Party's candidate" | q8 == "Democratic Party's candidate", qsupport := '1-D']
data_sept18[q7 == "Republican Party's candidate" | q8 == "Republican Party's candidate", qsupport := '2-R']
data_sept18[q8 == "(VOL) Other", qsupport := '3-O']
data_sept18[is.na(qsupport), qsupport := '4-DK/R']
data_sept18[, .(.N, pct = .N/nrow(data_sept18)), .(q7, q8, qsupport)][order(qsupport)]

calculate percentages by state

data_state = data_sept18[, .(y_dem = mean(qsupport == '1-D')
                             , y_rep = mean(qsupport == '2-R')
                             , y_other = mean(qsupport == '3-O')
                             , y_dkr = mean(qsupport == '4-DK/R')
                             , total_respondents = .N
                             # sample covariates
                             , age_under30 = mean(as.numeric(age) < 30)
                             , race_W = mean(racecmb == 'White')
                             ), by = sstate][order(sstate)]
head(data_state)
data_sept18[, .N, hh1][order(hh1)]
X = data_sept18[, .(state = as.character(sstate)
                    , sex_male = as.numeric(sex == 'Male')
                    , sex_female = as.numeric(sex == 'Female')
                    , age = as.numeric(age)
                    
                    , educ_postgrad = as.numeric(grepl("Postgraduate", educ))
                    , educ_bach = as.numeric(grepl("Four year", educ))
                    , educ_assoc = as.numeric(grepl("Two year associate|Sone college", educ))
                    , educ_highschool = as.numeric(grepl("High school graduate", educ))
                    , educ_none = as.numeric(grepl("Less than high school|High school incomplete", educ))
                    
                    , hisp = as.numeric(hisp == 'Yes')
                    , race_white = as.numeric(racecmb == 'White')
                    , race_black = as.numeric(racecmb == 'Black or African-American')
                    , race_asian = as.numeric(racecmb == 'Asian or Asian-American')
                    , race_mixedother = as.numeric(racecmb == 'Mixed Race' | racecmb == 'Or some other race')
                    
                    , relig_protestant = as.numeric(grepl("Protestant|Christian", relig))
                    , relig_catholic = as.numeric(grepl("Catholic", relig))
                    , relig_athiest = as.numeric(grepl("Athiest|Agnostic", relig))
                    , relig_jewish = as.numeric(grepl("Jewish", relig))
                    , relig_muslim = as.numeric(grepl("Muslim", relig))
                    , relig_LDS = as.numeric(grepl("Mormon", relig))
                    
                    , hh_n = ifelse(hh1 == '8 or more', 8, ifelse(hh1 == "Don't know/Refused", 2, as.numeric(hh1)))
                    
                    , income_under10 = as.numeric(income == 'Less than $10,000')
                    , income_10to20 = as.numeric(income == '10 to under $20,000')
                    , income_20to30 = as.numeric(income == '20 to under $30,000')
                    , income_30to40 = as.numeric(income == '30 to under $40,000')
                    , income_40to50 = as.numeric(income == '40 to under $50,000')
                    , income_50to75 = as.numeric(income == '50 to under $75,000')
                    , income_75to100 = as.numeric(income == '75 to under $100,000')
                    , income_100to150 = as.numeric(income == '100 to under $150,000')
                    , income_over150 = as.numeric(income == 'Over $150,000')
                    , income_refused = as.numeric(income == "(VOL) Don't know/Refused")
                    )]
# center and scale X
X_numeric = scale(X[, 2:ncol(X)])
# replace NA and NaN with 0
X_numeric = apply(X_numeric, 2, function(x) {
  x[is.na(x) | is.nan(x)] <- 0
  x
})
# create full covar matrix
X = data.frame(state = X$state, X_numeric)
##### Create test/training split
set.seed(123)
# train_ind = createDataPartition(X
#                                 , y = X$state
#                                 , p = .6, list = FALSE)
train_ind = sample.int(nrow(X), size = 1000, replace = F)
X_train = X[train_ind, ]
X_test = X[-train_ind, ]
Y_train = data_sept18[train_ind, .(state = as.character(sstate), y_dem = as.numeric(qsupport == '1-D'))]
Y_test = data_sept18[-train_ind, .(state = as.character(sstate), y_dem = as.numeric(qsupport == '1-D'))]
# create data that we actually observe
Y_train_agg = Y_train[, .(y_dem_pct = mean(y_dem), .N), by = state]
Y_test_agg = Y_test[, .(y_dem_pct = mean(y_dem), .N), by = state]
Y_train_agg

Mean embeddings

We observe bags of points \(\{x_j^i\}_{j = 1}^{N_i}\) from \(i = 1, \dots, n\), where \(x_j^i \in \mathbb{R}^p\). We have to embed the \(x_j^i\) in feature space and then take the mean. This means we can’t use the kernel trick which would take us directly from \(\mathbb{R}^p \times \mathbb{R}^p \to \mathbb{R}\). In order to calculate the coordinate of each observation in feature space, we convert each point to an explicit featurization: \(x_j^i \in \mathbb{R}\) is mapped to \[\phi(x_j^i) = [k(x_j^i, u_1), \dots, k(x_j^i, u_d)]^T \in \mathbb{R}^d\] where \(\mathbf{u} = \{u_l\}_{l = 1}^d\) are a set of landmark points. For now we’ll set \(d = 100\) and choose \(u_l\) using k-means clustering.

Specify landmark points \(\mathbf{u} = (u_1, \dots, u_k)\) as the centroids of k-means clustering

n_centers = 100
groups = kmeans(X_train[, -1], centers = n_centers)
u = as.matrix(groups$centers)
head(u)
     sex_male  sex_female          age educ_postgrad   educ_bach educ_assoc educ_highschool  educ_none       hisp race_white
1  0.90084435 -0.90084435  0.006395155   -0.35356593  1.65478397 -0.3815408      -0.4857439 -0.2324988 -0.3766247  0.3196829
2  0.23075063 -0.23075063  0.302619660   -0.35356593 -0.60396396 -0.3815408      -0.4857439  4.2986449 -0.3766247  0.5618264
3  0.09673189 -0.09673189 -0.314968403   -0.03553791  0.07366042  0.2186588      -0.4857439 -0.2324988  1.4415397 -1.7788945
4  0.90084435 -0.90084435 -0.718954580    2.82671430 -0.60396396 -0.3815408      -0.4857439 -0.2324988 -0.3766247  0.5618264
5  0.90084435 -0.90084435 -0.806020567   -0.35356593  0.52541001 -0.3815408      -0.4857439  2.0330730  1.1385123 -1.7788945
6 -1.10943680  1.10943680  0.587035216   -0.35356593 -0.15221437 -0.3815408       0.7011146 -0.2324988 -0.3766247  0.5618264
  race_black race_asian race_mixedother relig_protestant relig_catholic relig_athiest relig_jewish relig_muslim  relig_LDS
1 -0.2660408 -0.1546641      -0.3474161      -0.89153704      0.8273477    -0.2112912   -0.1658855  -0.08297413 -0.1447158
2 -0.2660408 -0.1546641      -0.3474161       0.67378425     -0.4866399    -0.2112912   -0.1658855  -0.08297413 -0.1447158
3 -0.2660408 -0.1546641       2.8767521      -0.08651466      0.7835481    -0.2112912   -0.1658855  -0.08297413 -0.1447158
4 -0.2660408 -0.1546641      -0.3474161       1.12101890     -0.4866399    -0.2112912   -0.1658855  -0.08297413 -0.1447158
5 -0.2660408  3.1536379       1.2646680      -0.89153704     -0.4866399    -0.2112912   -0.1658855  -0.08297413  6.9061577
6 -0.2660408 -0.1546641      -0.3474161       0.58433732     -0.1479231    -0.2112912   -0.1658855  -0.08297413 -0.1447158
         hh_n income_under10 income_10to20 income_20to30 income_30to40 income_40to50 income_50to75 income_75to100
1  0.20112327     -0.2053351    -0.2745375    -0.3101534    -0.2967138     -0.261095    -0.3815408    -0.36469037
2  0.02109454     -0.2053351    -0.2745375     0.8673554     1.3321941     -0.261095    -0.3815408    -0.01967046
3  0.04379381     -0.2053351    -0.2745375    -0.3101534    -0.2967138     -0.261095    -0.3815408     2.74048880
4  1.38305102     -0.2053351    -0.2745375    -0.3101534    -0.2967138     -0.261095    -0.3815408     2.74048880
5  1.95053289     -0.2053351    -0.2745375     1.4561097     1.5358076     -0.261095    -0.3815408    -0.36469037
6 -0.56908661     -0.2053351    -0.2745375     3.2223729    -0.2967138     -0.261095    -0.3815408    -0.36469037
  income_100to150 income_over150 income_refused
1               0              0     -0.2600654
2               0              0     -0.3666934
3               0              0     -0.3666934
4               0              0     -0.3666934
5               0              0     -0.3666934
6               0              0     -0.3666934

Create RBF kernel and get embeddings of training set \(\phi(x)\) using kernel and landmark points \[\phi(x_i) = [k(x_i, u_1), \dots, k(x_i, u_k)]\]

Choose scale parameter \(\sigma\) for RBF kernel - use median heuristic for now

rbf1 = rbfdot(sigma = 1)
K_sigma = kernelMatrix(rbf1, x = as.matrix(X_train[, -1]), y = u)
sigma = median(K_sigma)
ggplot(melt(K_sigma), aes(x = log(value))) + geom_histogram() + geom_vline(xintercept = log(sigma), color = 'red')

rbf = rbfdot(sigma = 0.02)
phi_x = kernelMatrix(rbf, x = as.matrix(X_train[, -1]), y = u)
phi_x = data.table(state = X_train$state, phi_x)
setnames(phi_x, c('state', paste0('u', 1:n_centers)))
head(phi_x)

Calculate mean embeddings for each state

mu_hat = phi_x[, lapply(.SD, mean), .SDcols = names(phi_x)[2:ncol(phi_x)], by = state][order(state)]
mu_hat

Model 1 - Frequentist

Normal LASSO regression on the kernel mean embeddings - no Bayesian treatment.

\[y_i = \beta^T\hat{\mu}_i + b\] Doesn’t account for uncertainty in the coefficients \(\beta\) or in the mean embeddings \(\hat{\mu}_i\), or observation error

Build the model

Use 5-fold CV kernel LASSO to estimate \(\beta\), \(b\) for \(\hat{y}_i = \beta^T\hat{\mu}_i + b\). Use the first LASSO to select non-zero coefs, then re-fit without penalty to accont for covariate shrinkage.

# fit initial lasso to find significant covars
fit_lambda = cv.glmnet(x = as.matrix(mu_hat[, -1, with = F])
                       , y = Y_train_agg[order(state), ]$y_dem_pct
                       , nfolds = 7)

# get indicies of non-0 covars (except intercept)
ind = which(coef(fit_lambda, s = 'lambda.min')[-1] != 0)

# re-fit LASSO only with non-0 covars
fit = glmnet(x = as.matrix(mu_hat[, ind + 1, with = F])
             , y = Y_train_agg[order(state), ]$y_dem_pct)

Predict

Get embeddings of test points and empirical means of bags of embedded test points. Use same landmark points \(u\) as for the training set.

# get embeddings of test points
phi_x_test = kernelMatrix(rbf, x = as.matrix(X_test[, -1]), y = u)
phi_x_test = data.table(state = X_test$state, phi_x_test)
setnames(phi_x_test, c('state', paste0('u', 1:n_centers)))

# calculate empirical means
mu_hat_test = phi_x_test[, lapply(.SD, mean), .SDcols = names(phi_x_test)[2:ncol(phi_x_test)], by = state][order(state)]
mu_hat_test_gender = cbind(phi_x_test, sex_female = as.numeric(X_test$sex_female > 0))[, lapply(.SD, mean), .SDcols = names(phi_x_test)[2:ncol(phi_x_test)], by = .(state, sex_female)][order(state, sex_female)]

First, test predictions at the state-level. We can also test at the state-gender level.

y_hat_agg = predict(fit, newx = as.matrix(mu_hat_test[, ind + 1, with = F]), s = 0)
y_hat_agg = cbind(Y_test[, .(y_dem_pct = mean(y_dem), .N), by = state][order(state)], y_hat_agg)

ggplot(y_hat_agg, aes(x = y_dem_pct, y = `1`, size = N)) + 
  geom_point() + 
  ylab("Predicted % Dem") +
  xlab("Actual % Dem") +
  ggtitle("Predicted by actual % Dem support across states")

y_hat_agg_gender = predict(fit, newx = as.matrix(mu_hat_test_gender[, ind + 1, with = F]), s = 0)
y_hat_agg_gender = cbind(cbind(Y_test, sex_female = as.numeric(X_test$sex_female > 0))[, .(y_dem_pct = mean(y_dem), .N), by = .(state, sex_female)][order(state, sex_female)], y_hat_agg_gender)

ggplot(y_hat_agg_gender, aes(x = y_dem_pct, y = `1`, size = N, color = as.factor(sex_female))) + 
  geom_point() + 
  geom_abline(intercept = 0, slope = 1) +
  ylab("Predicted % Dem") +
  xlab("Actual % Dem") +
  ggtitle("Predicted by actual % Dem support across states & genders")

print(paste0("MSE (state): ", mean((y_hat_agg[,`1`] - y_hat_agg$y_dem_pct)^2)))
print(paste0("MSE (state & gender): ", mean((y_hat_agg_gender[,`1`] - y_hat_agg_gender$y_dem_pct)^2)))

Predict at the individual-level

# predict on new (individual-level) data set of X_test embedded in feature space
y_hat = predict(fit, newx = as.matrix(phi_x_test[, ind + 1, with = F]), s= 0)  #no regularization
y_hat = cbind(Y_test, y_hat)

# calculate overall classification rate
class_rates = y_hat[, .(.N
          , pct_correct = mean((`1` > 0.5 & y_dem == 1) | (`1` <= 0.5 & y_dem == 0))
          , pct_false_neg = mean(`1` <= 0.5 & y_dem != 0)
          , pct_false_pos = mean(`1` > 0.5 & y_dem != 1)
          , mse = mean((y_dem - `1`)^2)
          )]
class_rates

Plot actual percent Dem by decile of the score

# get deciles of predicted score
y_hat[, pred_decile := cut(`1`, breaks = quantile(`1`, probs = seq(0,1,0.1)), labels = 1:10, include.lowest = T)]

#ggplot(y_hat, aes(x = `1`, color = as.factor(y_dem)), alpha = 0.2) + geom_density()

#plot actual pct of each decile that supports dems
ggplot(y_hat[, .(pct_y_dem = mean(y_dem)), by = pred_decile], aes(x = pred_decile, y = pct_y_dem)) + 
  geom_bar(stat = 'identity') +
  ggtitle("Pct Dem supporter by score decile")

Model 2 - Bayesian Linear Regression

Adding uncertainty: incorporate uncertainty of the coefficients \(\beta\)

We can interpret this in 2 ways: 1) Bayesian linear regression 2) Gaussian process regression conditioned on finite set of observed points.

  1. Bayesian linear regression \[\beta \sim \mathcal{N}(0, \Sigma_p)\\ y_i | \beta, \mathbf{x}_i \sim \mathcal{N}(\beta^T\hat{\mu}_i, \sigma^2)\]

For disitribution regression, we substitute in the empirical mean embeddings \(\hat{\mu}_i\) for \(x_i\).

  1. GP regression conditioned on observed points (weight space view) \[ f(\mathbf{x}) = \phi(\mathbf{x})^T\mathbf{w}\\ y = f(\mathbf{x}) + \epsilon\\ \epsilon \sim \mathcal{N}(0, \sigma_n^2)\\ \mathbf{w} \sim \mathcal{N}(0, \Sigma_p) \]

we can equivalently write \[ f \sim \mathcal{GP}(0, k(x, x'))\\ y_i|f(x_i) \sim \mathcal{N}(0, K + I\sigma^2) \] where \(k(x,x')\) is the kernel corresponding to the feature map \(\phi\)

For GPR, the feature map \(\phi(x_i)\) corresponds to the kernel that was used to embed \(x_i\) in feature space. Using landmark points \(\mathbf{u} = \{u_l\}_{l=1}^d\), this becomes \(\phi(x_i) = [k(x_i, u_1), \dots, k(x_i, u_l)]\)

We can interpret the weights \(\mathbf{w}\) of GP regression as the coefficients \(\beta\) in BLR. \(\Sigma_p\) is the prior covariance matrix for the weights/coefficients.

Fit GP

Now, we predict the GP at the bag-level


y_hat_agg = predict(fit, newx = as.matrix(mu_hat_test[, ind + 1, with = F]), s = 0)
y_hat_agg = cbind(Y_test[, .(y_dem_pct = mean(y_dem), .N), by = state][order(state)], y_hat_agg)

ggplot(y_hat_agg, aes(x = y_dem_pct, y = `1`, size = N)) + 
  geom_point() + 
  ylab("Predicted % Dem") +
  xlab("Actual % Dem") +
  ggtitle("Predicted by actual % Dem support across states")

And at the individual level

y_hat_gp = predict(fit_gp, newdata = phi_x_test[, -1])
y_hat_gp = cbind(Y_test, y_hat_gp)

y_hat_gp[, pred_decile := cut(V1, breaks = quantile(V1, probs = seq(0,1,0.1)), labels = 1:10, include.lowest = T)]

#plot actual pct of each decile that supports dems
ggplot(y_hat_gp[, .(pct_y_dem = mean(y_dem)), by = pred_decile], aes(x = pred_decile, y = pct_y_dem)) + 
  geom_bar(stat = 'identity') +
  ggtitle("Pct Dem supporter by score decile")

THINGS TO DO

“Unlabeled data can simply be used as centers” - Szabo 2016 - not entirely applicable here, we observe labels, just at the aggregate level

General questions

  • how much to focus on this specific example v. the more general question of “does dist reg work for individuals”?
  • What are the reasons that you’re skeptical that BDR will work for individuals? What are the shortcomings that you envision?

  • advantages to implementing base model as GPR v. BLR? GPR more flexible for non-Gaussian outcomes (i.e. multinomial)? Can combine kernels (i.e. mean embeddings + spatial component)? Downside is computation?

  • batching?

Choose data set:

  • census microdata with actual county-level election outcomes. pro: large (like actual voterfile), con: outcomes not at the individual-level
  • Pew survey data. pro: outcomes at the individual-level, con: it’s small
  • census microdata for covar data with survey data for outomes. Pro: large covar data, closer to real scenario, con: can only link at the state-level, still no individual-level outcome data
  • synthetic data. pro: can make it exactly what we want, con: effort to synthesize

** benefit of the larger data set - more realistic, we’d really need the landmark points when we don’t if only working with the survey data

** larger data also has more covariate data available

parameter tuning:

  • number of landmark points (how is this normally chosen?) - Leon’s code has 30 as the default for 1000 test, 1000 training (similar scale to data here)
  • Calculate landmark points with full data set or just training?
  • scale parameter \(\sigma\) for RBF kernel (right now just usiing median heuristic)

Choosing bag definitions

  • Since we have individual-level labels in the survey data, we can choose how we aggregate it (do we even need to aggregate it? YES because we need the strata definitions in order to link it back to the full file)
  • Bags based on demos instead of state (more specific to this poll than most we’d be dealing with)?

RBF Network v. Single RBF kernel

  • relationship between neural network implementation (re Leon’s code) and the kernel mean embedding? Just that we can exactly interpret kernel embedding as a neural network and maybe easier to implement/fast optimization of marginal log likelihood?

Benchmark:

  • Compare against lasso/custom-built logit or maybe NN that has access to the individual-level outcomes (more realistic benchmark than), just using a biased subset of the training observations (mimic asymmetric matching)
  • MRP - standard for EI in political applications
  • straight stratification using covariates observed in the survey

Implement full Bayesian treatment (uncertainty in mean embeddings as well as bag size)

  • flesh out full generative model

Structured observational uncertainty

  • heriarchical model?
---
title: "Bayesian DB Practice"
output: html_notebook
---


```{r setup, echo=F}
rm(list = ls())
library(data.table)
library(foreign)
library(kernlab)
library(MASS)  # for mvrnorm
library(ggplot2)
library(gridExtra)
library(glmnet)
library(caret)
setwd('~/github/bdr/')
```

## Overview

Bayesian distribution regression (BDR) is built on a few pieces of theory:

1. Kernel mean embeddings
2. Gaussian process regression (as the prior for the kernel mean embeddings)
3. Landmark points for dimensionality reduction (advantages of this over FastFood? both are reducing the dimensionality of the feature embedding so that we can calualate the empirical mean)

This notebook builds the theory of Bayesian Distribution Regression piece by piece:

1. Explore kernels from the `kernlab` package.  We'll later use the functions from this package to calculate mean embeddings for bags of observations $\{x_j^i\}_{j=1}^{N_i}$.

## Problem statement

We observe survey responses $(\mathbf{x}_i, y_i)$ for $i \in S$, where $\mathbf{x}_i \in \mathbb{R}^{V_s}$ is a vector of $v \in (1, \dots V_s)$ covariates (i.e. age, gender, education, income) and $y_i$ denotes who a respondent says they're likely to support in an upcoming political election. $S \subset (1, \dots, N)$ denotes the set of indicies in the population of size $N$ that were observed in the survey.  

We also observe a much larger set of $V_u >> V_s$ covariates $\mathbf{x}_i \in \mathbb{R}^{V_u}$, $V_s \subset V_u$, for all $i \in D$, where $D \subset N$.  The covariates $v \in V_u \setminus V_s$ are highly predictive of $y_i$.

We say that a survey respondent is "matched" to the population if $i \in M$ where $M = S \cap D$.  

The goal here is to estimate $\hat{y}_i \forall i \in D$.  Usually to do this, predictive models are built using only the "matched" data, or using respondents $i \in M$.  However, we would like to leverage $i \in S \setminus M$, for which we observe $y_i$ and $x_{i, 1}, \dots, x_{i, V_s}$ but not $\mathbf{x}_{i, \mathbf{v}}$ for $\mathbf{v} \in V_u \setminus V_s$.



Question:

- Is this better than predicting using the class proportion in each stratum defined by the covariates observed in the sample $p_s$?



## Explore Kernels from `kernlab` package

```{r kernels}
## create a RBF kernel function with sigma hyper-parameter 0.05
rbf = rbfdot(sigma = 1)

## create artificial data set
x <- matrix(rnorm(60), 6, 10)

## compute kernel matrix
kx <- kernelMatrix(rbf, x)  ## k_12 is equivalent to exp(- sum((x[1,] - x[2,])^2))
```


## Basic Gaussian process

A Gaussian process is a collection of random variables, any finite subset of which follows a multivariate normal distribution

*prior*: $f \sim \mathcal{GP}(\mathbf{0}, k(x, x'))$

We fix a finite set of $d$ points $\mathbf{s} = (s_1, \dots, s_d)$ in order to draw from the GP as a multivariate normal. Conditioning on $s$, this becomes $\mathbf{f} \sim \mathcal{N}(\mathbf{0}_d, \Sigma_d)$ where $\Sigma_d$ is the empirical covariance matrix of our points $s$.

We then ``observe" 5 points $\{x_i, y_i\}_{i = 1}^n$.  We assume that y is observed without noise such that $y = f(x)$.  In order to predict at a new point $x_\star$, we condiditon on the observed points $X$ and their associated outcomes $Y$, and arrive at the closed-form posterior predictive distribution of $f_\star$

*posterior*: $f_\star | x_\star, X, \mathbf{y} \sim \mathcal{N}(k_\star K^{-1}\mathbf{y}, k_{\star\star} - k_\star K^{-1}k_\star^T)$

where $K_{ij} = k(x_i, x_j)$ and $k_\star = [k(x_1, x_\star), \dots, k(x_n, x_\star)]$ and $k_{\star\star} = k(x_star, x_star)$

```{r simple-GP, fig.height=3, fig.width=6}
# define function to calculate empirical covariance matrix based on Gaussian kernel with length-scale l
calcSigma = function(x1, x2, l = 1){
  Sigma = matrix(nrow = length(x1), ncol = length(x2))
  
  for(i in 1:length(x1)){
    for(j in 1:length(x2)){
      Sigma[i,j] = exp(-1/2 * ((x1[i] - x2[j])/l)^2)
    }
  }
  return(Sigma)
}

# define the points that we want to fix for the function draw
s = seq(-5, 5, by = 0.1)

# calculate the empirical covariance of the points (basically 0)
Sigma = calcSigma(s, s)

# draw 5 samples from the GP prior at the fixed points
x_prior = t(mvrnorm(n = 5, mu = rep(0, length(s)), Sigma = Sigma))
x_draws = data.frame(cbind(s, x_prior))
x_draws = melt(x_draws, id = 's', value.name = 'f(x)')

# plot draws from the prior
plot_gp_prior = ggplot(x_draws, aes(x = s, y = `f(x)`, color = variable)) + geom_line() +
  ggtitle("Draws from GP Prior") + theme_minimal()

# specify points that we observe
x_obs = c(-4, -2.5, -1, 0, 4)
y_obs = c(-2, 0, 1, 2, -1)

# calculate kernel matricies needed for posterior evaluation
K_xx = calcSigma(x_obs, x_obs)
K_xstarx = calcSigma(s, x_obs)
K_xxstar = calcSigma(x_obs, s)
K_xstarxstar = calcSigma(s, s)

# calculate posterior mean and variance
mu = K_xstarx %*% solve(K_xx) %*% y_obs
Sigma_star = K_xstarxstar - K_xstarx %*% solve(K_xx) %*% K_xxstar

# draw 5 samples from the posterior at fixed eval points s
x_post = t(mvrnorm(5, mu = mu, Sigma = Sigma_star))
x_post = data.frame(cbind(s, x_post))
x_post = melt(x_post, id = 's', value.name = 'f_star')

# calculate 95% confidence bands for posterior
f_star_covar = data.frame(cbind(s, ub = mu + 2 * sqrt(diag(Sigma_star)), lb = mu - 2 * sqrt(diag(Sigma_star))))

# plot draws from posterior
plot_gp_post = ggplot() + 
  geom_ribbon(data = f_star_covar, aes(x = s, ymin = V3, ymax = V2), fill = "grey", alpha = 0.5) +
  geom_line(data = x_post, aes(x = s, y = f_star, color = variable)) +
  #geom_point(aes(x = x_obs, y = y_obs)) +
  ggtitle("Draws from GP Posterior") +
  theme_minimal()
  
grid.arrange(plot_gp_prior, plot_gp_post, nrow = 1)
```


## Try out BDR on survey data

*Goal*: Estimate individual-level support using BDR on state-level support outcomes 

### Data

```{r load-data, echo = F, message= F}

# note: geo variables that begin with 's' are self-reported (I think based on the readme) - more accurate

data_sept18 = data.table(read.spss('data/Sept18/Sept18 public.sav', to.data.frame = T), stringsAsFactors = F)
data_sept18
```

Individual-level survey responses (n = `nrow(data_sept18`) from September 2018 Pew Research survey (https://www.people-press.org/dataset/september-2018-political-survey/).



#### Recode survey responses

Recode the survey responses - combine q7 (strong) and q8 (leaners) to get full support response
```{r recode-data, message = F}
## support 
data_sept18[, .N, .(q7, q8)]
data_sept18[, qsupport := NULL]
data_sept18[q7 == "Democratic Party's candidate" | q8 == "Democratic Party's candidate", qsupport := '1-D']
data_sept18[q7 == "Republican Party's candidate" | q8 == "Republican Party's candidate", qsupport := '2-R']
data_sept18[q8 == "(VOL) Other", qsupport := '3-O']
data_sept18[is.na(qsupport), qsupport := '4-DK/R']

data_sept18[, .(.N, pct = .N/nrow(data_sept18)), .(q7, q8, qsupport)][order(qsupport)]
```

# calculate percentages by state
```{r}
data_state = data_sept18[, .(y_dem = mean(qsupport == '1-D')
                             , y_rep = mean(qsupport == '2-R')
                             , y_other = mean(qsupport == '3-O')
                             , y_dkr = mean(qsupport == '4-DK/R')
                             , total_respondents = .N
                             # sample covariates
                             , age_under30 = mean(as.numeric(age) < 30)
                             , race_W = mean(racecmb == 'White')
                             ), by = sstate][order(sstate)]

head(data_state)
```


```{r}
data_sept18[, .N, hh1][order(hh1)]

X = data_sept18[, .(state = as.character(sstate)
                    , sex_male = as.numeric(sex == 'Male')
                    , sex_female = as.numeric(sex == 'Female')
                    , age = as.numeric(age)
                    
                    , educ_postgrad = as.numeric(grepl("Postgraduate", educ))
                    , educ_bach = as.numeric(grepl("Four year", educ))
                    , educ_assoc = as.numeric(grepl("Two year associate|Sone college", educ))
                    , educ_highschool = as.numeric(grepl("High school graduate", educ))
                    , educ_none = as.numeric(grepl("Less than high school|High school incomplete", educ))
                    
                    , hisp = as.numeric(hisp == 'Yes')
                    , race_white = as.numeric(racecmb == 'White')
                    , race_black = as.numeric(racecmb == 'Black or African-American')
                    , race_asian = as.numeric(racecmb == 'Asian or Asian-American')
                    , race_mixedother = as.numeric(racecmb == 'Mixed Race' | racecmb == 'Or some other race')
                    
                    , relig_protestant = as.numeric(grepl("Protestant|Christian", relig))
                    , relig_catholic = as.numeric(grepl("Catholic", relig))
                    , relig_athiest = as.numeric(grepl("Athiest|Agnostic", relig))
                    , relig_jewish = as.numeric(grepl("Jewish", relig))
                    , relig_muslim = as.numeric(grepl("Muslim", relig))
                    , relig_LDS = as.numeric(grepl("Mormon", relig))
                    
                    , hh_n = ifelse(hh1 == '8 or more', 8, ifelse(hh1 == "Don't know/Refused", 2, as.numeric(hh1)))
                    
                    , income_under10 = as.numeric(income == 'Less than $10,000')
                    , income_10to20 = as.numeric(income == '10 to under $20,000')
                    , income_20to30 = as.numeric(income == '20 to under $30,000')
                    , income_30to40 = as.numeric(income == '30 to under $40,000')
                    , income_40to50 = as.numeric(income == '40 to under $50,000')
                    , income_50to75 = as.numeric(income == '50 to under $75,000')
                    , income_75to100 = as.numeric(income == '75 to under $100,000')
                    , income_100to150 = as.numeric(income == '100 to under $150,000')
                    , income_over150 = as.numeric(income == 'Over $150,000')
                    , income_refused = as.numeric(income == "(VOL) Don't know/Refused")
                    )]

# center and scale X
X_numeric = scale(X[, 2:ncol(X)])

# replace NA and NaN with 0
X_numeric = apply(X_numeric, 2, function(x) {
  x[is.na(x) | is.nan(x)] <- 0
  x
})
# create full covar matrix
X = data.frame(state = X$state, X_numeric)


##### Create test/training split
set.seed(123)
# train_ind = createDataPartition(X
#                                 , y = X$state
#                                 , p = .6, list = FALSE)

train_ind = sample.int(nrow(X), size = 1000, replace = F)

X_train = X[train_ind, ]
X_test = X[-train_ind, ]

Y_train = data_sept18[train_ind, .(state = as.character(sstate), y_dem = as.numeric(qsupport == '1-D'))]
Y_test = data_sept18[-train_ind, .(state = as.character(sstate), y_dem = as.numeric(qsupport == '1-D'))]

# create data that we actually observe
Y_train_agg = Y_train[, .(y_dem_pct = mean(y_dem), .N), by = state]
Y_test_agg = Y_test[, .(y_dem_pct = mean(y_dem), .N), by = state]

Y_train_agg
```

## Mean embeddings

We observe bags of points $\{x_j^i\}_{j = 1}^{N_i}$ from $i = 1, \dots, n$, where $x_j^i \in \mathbb{R}^p$.  We have to embed the $x_j^i$ in feature space and then take the mean.  This means we can't use the kernel trick which would take us directly from $\mathbb{R}^p \times \mathbb{R}^p \to \mathbb{R}$.  In order to calculate the coordinate of each observation in feature space, we convert each point to an *explicit* featurization:  $x_j^i \in \mathbb{R}$ is mapped to $$\phi(x_j^i) = [k(x_j^i, u_1), \dots, k(x_j^i, u_d)]^T \in \mathbb{R}^d$$
where $\mathbf{u} = \{u_l\}_{l = 1}^d$ are a set of landmark points.  For now we'll set $d = 100$ and choose $u_l$ using k-means clustering.


Specify landmark points $\mathbf{u} = (u_1, \dots, u_k)$ as the centroids of k-means clustering

```{r}
n_centers = 100
groups = kmeans(X_train[, -1], centers = n_centers)
u = as.matrix(groups$centers)
head(u)
```


Create RBF kernel and get embeddings of training set $\phi(x)$ using kernel and landmark points
$$\phi(x_i) = [k(x_i, u_1), \dots, k(x_i, u_k)]$$

Choose scale parameter $\sigma$ for RBF kernel - use median heuristic for now
```{r}
rbf1 = rbfdot(sigma = 1)

K_sigma = kernelMatrix(rbf1, x = as.matrix(X_train[, -1]), y = u)
sigma = median(K_sigma)

ggplot(melt(K_sigma), aes(x = log(value))) + geom_histogram() + geom_vline(xintercept = log(sigma), color = 'red')
```


```{r}
rbf = rbfdot(sigma = 0.02)
phi_x = kernelMatrix(rbf, x = as.matrix(X_train[, -1]), y = u)
phi_x = data.table(state = X_train$state, phi_x)
setnames(phi_x, c('state', paste0('u', 1:n_centers)))

head(phi_x)
```

Calculate mean embeddings for each state

```{r}
mu_hat = phi_x[, lapply(.SD, mean), .SDcols = names(phi_x)[2:ncol(phi_x)], by = state][order(state)]
mu_hat
```

### Model 1 - Frequentist
Normal LASSO regression on the kernel mean embeddings - no Bayesian treatment.

$$y_i = \beta^T\hat{\mu}_i + b$$
Doesn't account for uncertainty in the coefficients $\beta$ or in the mean embeddings $\hat{\mu}_i$, or observation error

#### Build the model

Use 5-fold CV kernel LASSO to estimate $\beta$, $b$ for $\hat{y}_i = \beta^T\hat{\mu}_i + b$.  Use the first LASSO to select non-zero coefs, then re-fit without penalty to accont for covariate shrinkage.

```{r}
# fit initial lasso to find significant covars
fit_lambda = cv.glmnet(x = as.matrix(mu_hat[, -1, with = F])
                       , y = Y_train_agg[order(state), ]$y_dem_pct
                       , nfolds = 7)

# get indicies of non-0 covars (except intercept)
ind = which(coef(fit_lambda, s = 'lambda.min')[-1] != 0)

# re-fit LASSO only with non-0 covars
fit = glmnet(x = as.matrix(mu_hat[, ind + 1, with = F])
             , y = Y_train_agg[order(state), ]$y_dem_pct)
```


#### Predict

Get embeddings of test points and empirical means of bags of embedded test points.  Use same landmark points $u$ as for the training set.
```{r}
# get embeddings of test points
phi_x_test = kernelMatrix(rbf, x = as.matrix(X_test[, -1]), y = u)
phi_x_test = data.table(state = X_test$state, phi_x_test)
setnames(phi_x_test, c('state', paste0('u', 1:n_centers)))

# calculate empirical means
mu_hat_test = phi_x_test[, lapply(.SD, mean), .SDcols = names(phi_x_test)[2:ncol(phi_x_test)], by = state][order(state)]
mu_hat_test_gender = cbind(phi_x_test, sex_female = as.numeric(X_test$sex_female > 0))[, lapply(.SD, mean), .SDcols = names(phi_x_test)[2:ncol(phi_x_test)], by = .(state, sex_female)][order(state, sex_female)]
```


First, test predictions at the state-level.  We can also test at the state-gender level.
```{r}
y_hat_agg = predict(fit, newx = as.matrix(mu_hat_test[, ind + 1, with = F]), s = 0)
y_hat_agg = cbind(Y_test[, .(y_dem_pct = mean(y_dem), .N), by = state][order(state)], y_hat_agg)

ggplot(y_hat_agg, aes(x = y_dem_pct, y = `1`, size = N)) + 
  geom_point() + 
  ylab("Predicted % Dem") +
  xlab("Actual % Dem") +
  ggtitle("Predicted by actual % Dem support across states")

y_hat_agg_gender = predict(fit, newx = as.matrix(mu_hat_test_gender[, ind + 1, with = F]), s = 0)
y_hat_agg_gender = cbind(cbind(Y_test, sex_female = as.numeric(X_test$sex_female > 0))[, .(y_dem_pct = mean(y_dem), .N), by = .(state, sex_female)][order(state, sex_female)], y_hat_agg_gender)

ggplot(y_hat_agg_gender, aes(x = y_dem_pct, y = `1`, size = N, color = as.factor(sex_female))) + 
  geom_point() + 
  geom_abline(intercept = 0, slope = 1) +
  ylab("Predicted % Dem") +
  xlab("Actual % Dem") +
  ggtitle("Predicted by actual % Dem support across states & genders")

print(paste0("MSE (state): ", mean((y_hat_agg[,`1`] - y_hat_agg$y_dem_pct)^2)))
print(paste0("MSE (state & gender): ", mean((y_hat_agg_gender[,`1`] - y_hat_agg_gender$y_dem_pct)^2)))

```

Predict at the individual-level
```{r}
# predict on new (individual-level) data set of X_test embedded in feature space
y_hat = predict(fit, newx = as.matrix(phi_x_test[, ind + 1, with = F]), s= 0)  #no regularization
y_hat = cbind(Y_test, y_hat)

# calculate overall classification rate
class_rates = y_hat[, .(.N
          , pct_correct = mean((`1` > 0.5 & y_dem == 1) | (`1` <= 0.5 & y_dem == 0))
          , pct_false_neg = mean(`1` <= 0.5 & y_dem != 0)
          , pct_false_pos = mean(`1` > 0.5 & y_dem != 1)
          , mse = mean((y_dem - `1`)^2)
          )]
class_rates
```

Plot actual percent Dem by decile of the score
```{r}
# get deciles of predicted score
y_hat[, pred_decile := cut(`1`, breaks = quantile(`1`, probs = seq(0,1,0.1)), labels = 1:10, include.lowest = T)]

#ggplot(y_hat, aes(x = `1`, color = as.factor(y_dem)), alpha = 0.2) + geom_density()

#plot actual pct of each decile that supports dems
ggplot(y_hat[, .(pct_y_dem = mean(y_dem)), by = pred_decile], aes(x = pred_decile, y = pct_y_dem)) + 
  geom_bar(stat = 'identity') +
  ggtitle("Pct Dem supporter by score decile")
```



## Model 2 - Bayesian Linear Regression

Adding uncertainty: incorporate uncertainty of the coefficients $\beta$

We can interpret this in 2 ways: 1) Bayesian linear regression 2) Gaussian process regression conditioned on finite set of observed points.

1) Bayesian linear regression
$$\beta \sim \mathcal{N}(0, \Sigma_p)\\
y_i | \beta, \mathbf{x}_i \sim \mathcal{N}(\beta^T\hat{\mu}_i, \sigma^2)$$

For disitribution regression, we substitute in the empirical mean embeddings $\hat{\mu}_i$ for $x_i$.

2) GP regression conditioned on observed points (weight space view)
$$
f(\mathbf{x}) = \phi(\mathbf{x})^T\mathbf{w}\\
y = f(\mathbf{x}) + \epsilon\\
\epsilon \sim \mathcal{N}(0, \sigma_n^2)\\
\mathbf{w} \sim \mathcal{N}(0, \Sigma_p)
$$

we can equivalently write
$$
f \sim \mathcal{GP}(0, k(x, x'))\\
y_i|f(x_i) \sim \mathcal{N}(0, K + I\sigma^2)
$$
where $k(x,x')$ is the kernel corresponding to the feature map $\phi$

For GPR, the feature map $\phi(x_i)$ corresponds to the kernel that was used to embed $x_i$ in feature space.  Using landmark points $\mathbf{u} = \{u_l\}_{l=1}^d$, this becomes $\phi(x_i) = [k(x_i, u_1), \dots, k(x_i, u_l)]$

We can interpret the weights $\mathbf{w}$ of GP regression as the coefficients $\beta$ in BLR.  $\Sigma_p$ is the prior covariance matrix for the weights/coefficients.

Fit GP
```{r}
# observations = mu_hat
# outcomes = Y_train_agg$y_dem_pct

fit_gp = gausspr(x = as.matrix(mu_hat[, -1]), y = Y_train_agg[order(state),]$y_dem_pct, kernel = 'vanilladot')
fit_gp

#fit_gp@alpha
```

Now, we predict the GP at the bag-level
```{r}

y_hat_agg = predict(fit, newx = as.matrix(mu_hat_test[, ind + 1, with = F]), s = 0)
y_hat_agg = cbind(Y_test[, .(y_dem_pct = mean(y_dem), .N), by = state][order(state)], y_hat_agg)

ggplot(y_hat_agg, aes(x = y_dem_pct, y = `1`, size = N)) + 
  geom_point() + 
  ylab("Predicted % Dem") +
  xlab("Actual % Dem") +
  ggtitle("Predicted by actual % Dem support across states")
```

And at the individual level
```{r}
y_hat_gp = predict(fit_gp, newdata = phi_x_test[, -1])
y_hat_gp = cbind(Y_test, y_hat_gp)

y_hat_gp[, pred_decile := cut(V1, breaks = quantile(V1, probs = seq(0,1,0.1)), labels = 1:10, include.lowest = T)]

#plot actual pct of each decile that supports dems
ggplot(y_hat_gp[, .(pct_y_dem = mean(y_dem)), by = pred_decile], aes(x = pred_decile, y = pct_y_dem)) + 
  geom_bar(stat = 'identity') +
  ggtitle("Pct Dem supporter by score decile")
```









## THINGS TO DO

"Unlabeled data can simply be used as centers" - Szabo 2016
- not entirely applicable here, we observe labels, just at the aggregate level

*General questions*

- how much to focus on this specific example v. the more general question of "does dist reg work for individuals"?
- What are the reasons that you're skeptical that BDR will work for individuals?  What are the shortcomings that you envision?

- advantages to implementing base model as GPR v. BLR?  GPR more flexible for non-Gaussian outcomes (i.e. multinomial)?  Can combine kernels (i.e. mean embeddings + spatial component)?  Downside is computation?

- batching?

*Choose data set*:

- census microdata with actual county-level election outcomes. pro: large (like actual voterfile), con: outcomes not at the individual-level
- Pew survey data. pro: outcomes at the individual-level, con: it's small
- census microdata for covar data with survey data for outomes.  Pro: large covar data, closer to real scenario, con: can only link at the state-level, still no individual-level outcome data
- synthetic data. pro: can make it exactly what we want, con: effort to synthesize

** benefit of the larger data set - more realistic, we'd really need the landmark points when we don't if only working with the survey data

** larger data also has more covariate data available


*parameter tuning*:

- number of landmark points (how is this normally chosen?) - Leon's code has 30 as the default for 1000 test, 1000 training (similar scale to data here)
- Calculate landmark points with full data set or just training?
- scale parameter $\sigma$ for RBF kernel (right now just usiing median heuristic)

*Choosing bag definitions*

- Since we have individual-level labels in the survey data, we can choose how we aggregate it (do we even need to aggregate it? YES because we need the strata definitions in order to link it back to the full file)
- Bags based on demos instead of state (more specific to this poll than most we'd be dealing with)?

*RBF Network v. Single RBF kernel*

- relationship between neural network implementation (re Leon's code) and the kernel mean embedding? Just that we can exactly interpret kernel embedding as a neural network and maybe easier to implement/fast optimization of marginal log likelihood?

*Benchmark*: 

- Compare against lasso/custom-built logit or maybe NN that has access to the individual-level outcomes (more realistic benchmark than), just using a biased subset of the training observations (mimic asymmetric matching)
- MRP - standard for EI in political applications
- straight stratification using covariates observed in the survey


Implement *full Bayesian treatment* (uncertainty in mean embeddings as well as bag size)

- flesh out full generative model

*Structured observational uncertainty*

- heriarchical model?










